106 research outputs found
Recommended from our members
How well do CMIP5 climate simulations replicate historical trends and patterns of meteorological droughts?
Assessing the uncertainties and understanding the deficiencies of climate models are fundamental to developing adaptation strategies. The objective of this study is to understand how well Coupled Model Intercomparison-Phase 5 (CMIP5) climate model simulations replicate ground-based observations of continental drought areas and their trends. The CMIP5 multimodel ensemble encompasses the Climatic Research Unit (CRU) ground-based observations of area under drought at all time steps. However, most model members overestimate the areas under extreme drought, particularly in the Southern Hemisphere (SH). Furthermore, the results show that the time series of observations and CMIP5 simulations of areas under drought exhibit more variability in the SH than in the Northern Hemisphere (NH). The trend analysis of areas under drought reveals that the observational data exhibit a significant positive trend at the significance level of 0.05 over all land areas. The observed trend is reproduced by about three-fourths of the CMIP5 models when considering total land areas in drought. While models are generally consistent with observations at a global (or hemispheric) scale, most models do not agree with observed regional drying and wetting trends. Over many regions, at most 40% of the CMIP5 models are in agreement with the trends of CRU observations. The drying/wetting trends calculated using the 3 months Standardized Precipitation Index (SPI) values show better agreement with the corresponding CRU values than with the observed annual mean precipitation rates. Pixel-scale evaluation of CMIP5 models indicates that no single model demonstrates an overall superior performance relative to the other models
A new low-complexity patch-based image super-resolution
In this study, a novel single image super-resolution (SR) method, which uses a generated dictionary from pairs of high-resolution (HR) images and their corresponding low-resolution (LR) representations, is proposed. First, HR and LR dictionaries are created by dividing HR and LR images into patches Afterwards, when performing SR, the distance between every patch of the input LR image and those of available LR patches in the LR dictionary are calculated. The minimum distance between the input LR patch and those in the LR dictionary is taken, and its counterpart from the HR dictionary will be passed through an illumination enhancement process resulting in consistency of illumination between neighbour patches. This process is applied to all patches of the LR image. Finally, in order to remove the blocking effect caused by merging the patches, an average of the obtained HR image and the interpolated image is calculated. Furthermore, it is shown that the stabe of dictionaries is reducible to a great degree. The speed of the system is improved by 62.5%. The quantitative and qualitative analyses of the experimental results show the superiority of the proposed technique over the conventional and state-of-the-art methods
Deep Burst Denoising
Noise is an inherent issue of low-light image capture, one which is
exacerbated on mobile devices due to their narrow apertures and small sensors.
One strategy for mitigating noise in a low-light situation is to increase the
shutter time of the camera, thus allowing each photosite to integrate more
light and decrease noise variance. However, there are two downsides of long
exposures: (a) bright regions can exceed the sensor range, and (b) camera and
scene motion will result in blurred images. Another way of gathering more light
is to capture multiple short (thus noisy) frames in a "burst" and intelligently
integrate the content, thus avoiding the above downsides. In this paper, we use
the burst-capture strategy and implement the intelligent integration via a
recurrent fully convolutional deep neural net (CNN). We build our novel,
multiframe architecture to be a simple addition to any single frame denoising
model, and design to handle an arbitrary number of noisy input frames. We show
that it achieves state of the art denoising results on our burst dataset,
improving on the best published multi-frame techniques, such as VBM4D and
FlexISP. Finally, we explore other applications of image enhancement by
integrating content from multiple frames and demonstrate that our DNN
architecture generalizes well to image super-resolution
Improved interpolation kernels for super resolution algorithms
Super resolution (SR) algorithms are widely used in forensics investigations to enhance the resolution of images captured by surveillance cameras. Such algorithms usually use a common interpolation algorithm to generate an initial guess for the desired high resolution (HR) image. This initial guess is usually tuned through different methods, like learning-based or fusion-based methods, to converge the initial guess towards the desired HR output. In this work, it is shown that SR algorithms can result in better performance if more sophisticated kernels than the simple conventional ones are used for producing the initial guess. The contribution of this work is to introduce such a set of kernels which can be used in the context of SR. The quantitative and qualitative results on many natural, facial and iris images show the superiority of the generated HR images over two state-of-the-art SR algorithms when their original interpolation kernel is replaced by the ones introduced in this work
Super-Resolved Enhancement of a Single Image and Its Application in Cardiac MRI
Super-resolved image enhancement is of great importance in medical imaging. Conventional methods often require multiple low resolution (LR) images from different views of the same object or learning from large amount of training datasets to achieve success. However, in real clinical environments, these prerequisites are rarely fulfilled. In this paper, we present a self-learning based method to perform superresolution (SR) from a single LR input. The mappings between the given LR image and its downsampled versions are modeled using support vector regression on features extracted from sparse coded dictionaries, coupled with dual-tree complex wavelet transform based denoising. We demonstrate the efficacy of our method in application of cardiac MRI enhancement. Both quantitative and qualitative results show that our SR method is able to preserve fine textural details that can be corrupted by noise, and therefore can maintain crucial diagnostic information
Simple, Accurate, and Robust Nonparametric Blind Super-Resolution
This paper proposes a simple, accurate, and robust approach to single image
nonparametric blind Super-Resolution (SR). This task is formulated as a
functional to be minimized with respect to both an intermediate super-resolved
image and a nonparametric blur-kernel. The proposed approach includes a
convolution consistency constraint which uses a non-blind learning-based SR
result to better guide the estimation process. Another key component is the
unnatural bi-l0-l2-norm regularization imposed on the super-resolved, sharp
image and the blur-kernel, which is shown to be quite beneficial for estimating
the blur-kernel accurately. The numerical optimization is implemented by
coupling the splitting augmented Lagrangian and the conjugate gradient (CG).
Using the pre-estimated blur-kernel, we finally reconstruct the SR image by a
very simple non-blind SR method that uses a natural image prior. The proposed
approach is demonstrated to achieve better performance than the recent method
by Michaeli and Irani [2] in both terms of the kernel estimation accuracy and
image SR quality
Projected changes of rainfall seasonality and dry spells in a high greenhouse gas emissions scenario
In this diagnostic study we analyze changes of rainfall seasonality and dry spells by the end of the twenty-first century under the most extreme IPCC5 emission scenario (RCP8.5) as projected by twenty-four coupled climate models contributing to Coupled Model Intercomparison Project 5 (CMIP5). We use estimates of the centroid of the monthly rainfall distribution as an index of the rainfall timing and a threshold-independent, information theory-based quantity such as relative entropy (RE) to quantify the concentration of annual rainfall and the number of dry months and to build a monsoon dimensionless seasonality index (DSI). The RE is projected to increase, with high inter-model agreement over Mediterranean-type regions---southern Europe, northern Africa and southern Australia---and areas of South and Central America, implying an increase in the number of dry days up to 1Â month by the end of the twenty-first century. Positive RE changes are also projected over the monsoon regions of southern Africa and North America, South America. These trends are consistent with a shortening of the wet season associated with a more prolonged pre-monsoonal dry period. The extent of the global monsoon region, characterized by large DSI, is projected to remain substantially unaltered. Centroid analysis shows that most of CMIP5 projections suggest that the monsoonal annual rainfall distribution is expected to change from early to late in the course of the hydrological year by the end of the twenty-first century and particularly after year 2050. This trend is particularly evident over northern Africa, southern Africa and western Mexico, where more than 90% of the models project a delay of the rainfall centroid from a few days up to 2Â weeks. Over the remaining monsoonal regions, there is little inter-model agreement in terms of centroid changes
- …